Graph Based Heuristics : Beating the LRU Standard ?
نویسندگان
چکیده
In this paper we devise new paging heuristics motivated by the access graph model of paging 2]. Unlike the access graph model 2, 7, 4] and the related Markov paging model 8] our heuristics are truly on-line in that we do not assume any prior knowledge of the program just about to be executed. The Least Recently Used heuristic for paging is remarkably good, and is known experimentally to be superior to many of the suggested alternatives on real program traces 17]. Experiments we've performed suggest that our heuristics beat LRU fairly consistently, over a wide range of cache sizes and programs. The number of page faults can be as low as 1/2 the number of page faults for LRU and is on average 7 to 9 percent less than the number of faults for LRU. (Depending on how this average is computed.) We have built a program tracer that gives the page access sequence for real program executions of 200 { 3,300 thousand page access requests, and our simulations are based on 25 of these program traces. Overall, we have performed several thousand such simulations. While we have no real evidence to suggest that the programs we've traced are typical in any sense, we have made use of an experimental \protocol" designed to avoid experimenter bias. We strongly emphasize that our results are only preliminary and that much further work needs to be done.
منابع مشابه
A New Case for Skewed-Associativity
Skewed-associative caches have been shown to statisticaly exhibit lower miss ratios than set-associative caches at equal associativity degrees. But for L1 caches, the performance of a four-way set-associative cache is known to be quite close to that of a 16-way set-associative cache. In this paper, we rst show that BTBs and L2 caches featuring long line size may beneet a lot from a high associa...
متن کاملA comparative performance of gray level image thresholding using normalized graph cut based standard S membership function
In this research paper, we use a normalized graph cut measure as a thresholding principle to separate an object from the background based on the standard S membership function. The implementation of the proposed algorithm known as fuzzy normalized graph cut method. This proposed algorithm compared with the fuzzy entropy method [25], Kittler [11], Rosin [21], Sauvola [23] and Wolf [33] method. M...
متن کاملLRU-SP: A Size-Adjusted and Popularity-Aware LRU Replacement Algorithm for Web Caching
This paper presents LRU-SP, a size-adjusted and popularity-aware extension to Least Recently Used (LRU) for caching web objects. The standard LRU, focusing on recently used and equal sized objects, is not suitable for the web context because web objects vary dramatically in size and the recently accessed objects may possibly differ from popular ones. LRU-SP is built on two LRU extensions, namel...
متن کاملTheoretical Evidence for the Superiority of LRU-2 over LRU for the Paging Problem
The paging algorithm LRU-2 was proposed for use in database disk buffering and shown experimentally to perform better than LRU [O’Neil, O’Neil, and Weikum, 1993]. We compare LRU-2 and LRU theoretically, using both the standard competitive analysis and the newer relative worst order analysis. The competitive ratio for LRU-2 is shown to be 2k for cache size k, which is worse than LRU’s competitiv...
متن کاملIntelligent Naïve Bayes-based approaches for Web proxy caching
Web proxy caching is one of the most successful solutions for improving the performance of Web-based systems. In Web proxy caching, the popular Web objects that are likely to be revisited in the near future are stored on the proxy server, which plays the key roles between users and Web sites in reducing the response time of user requests and saving the network bandwidth. However, the difficulty...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1997